216 research outputs found

    Skeleton based action recognition using translation-scale invariant image mapping and multi-scale deep cnn

    Full text link
    This paper presents an image classification based approach for skeleton-based video action recognition problem. Firstly, A dataset independent translation-scale invariant image mapping method is proposed, which transformes the skeleton videos to colour images, named skeleton-images. Secondly, A multi-scale deep convolutional neural network (CNN) architecture is proposed which could be built and fine-tuned on the powerful pre-trained CNNs, e.g., AlexNet, VGGNet, ResNet etal.. Even though the skeleton-images are very different from natural images, the fine-tune strategy still works well. At last, we prove that our method could also work well on 2D skeleton video data. We achieve the state-of-the-art results on the popular benchmard datasets e.g. NTU RGB+D, UTD-MHAD, MSRC-12, and G3D. Especially on the largest and challenge NTU RGB+D, UTD-MHAD, and MSRC-12 dataset, our method outperforms other methods by a large margion, which proves the efficacy of the proposed method

    Integrated Deep and Shallow Networks for Salient Object Detection

    Full text link
    Deep convolutional neural network (CNN) based salient object detection methods have achieved state-of-the-art performance and outperform those unsupervised methods with a wide margin. In this paper, we propose to integrate deep and unsupervised saliency for salient object detection under a unified framework. Specifically, our method takes results of unsupervised saliency (Robust Background Detection, RBD) and normalized color images as inputs, and directly learns an end-to-end mapping between inputs and the corresponding saliency maps. The color images are fed into a Fully Convolutional Neural Networks (FCNN) adapted from semantic segmentation to exploit high-level semantic cues for salient object detection. Then the results from deep FCNN and RBD are concatenated to feed into a shallow network to map the concatenated feature maps to saliency maps. Finally, to obtain a spatially consistent saliency map with sharp object boundaries, we fuse superpixel level saliency map at multi-scale. Extensive experimental results on 8 benchmark datasets demonstrate that the proposed method outperforms the state-of-the-art approaches with a margin.Comment: Accepted by IEEE International Conference on Image Processing (ICIP) 201

    Foreword to the special Issue on Hyperspectral remote sensing and imaging spectroscopy

    Get PDF
    The twenty six papers in this special issue focus on the technologies of hyperspectral remote sensing (HRS)and imaging spectroscopy. HRS has emerged as a powerful tool to understand phenomena at local and global scales by virtue of imaging through a diverse range of platforms, including terrestrial in-situ imaging platforms, unmanned and manned aerial vehicles, and satellite platforms. By virtue of imaging over a wide range of spectral wavelengths, it is possible to characterize object specific properties very accurately. As a result, hyperspectral imaging (also known as imaging spectroscopy) has gained popularity for a wide variety of applications, including environment monitoring, precision agriculture, mineralogy, forestry, urban planning, and defense applications. The increased analysis capability comes at a cost—there are a variety of challenges that must be overcome for robust image analysis of such data, including high dimensionality, limited sample size for training supervised models, noise and atmospheric affects, mixed pixels, etc. The papers in this issue represent some of the recent developments in image analysis algorithms and unique applications of hyperspectral imaging data
    • …
    corecore